# Multi-task Pre-training

Tarsier 34b
Apache-2.0
Tarsier-34b is an open-source large-scale video-language model focused on generating high-quality video captions and achieving leading results in multiple public benchmarks.
Video-to-Text Transformers
T
omni-research
103
17
All MiniLM L12 V2 GGUF
Apache-2.0
A lightweight sentence embedding model based on BERT for efficient calculation of sentence similarity
Text Embedding English
A
leliuga
1,169
4
Nepberta
This is a pre-trained language model specifically for Nepali, with research results published in the 2nd Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the 12th International Joint Conference on Natural Language Processing (2022).
Large Language Model Transformers
N
NepBERTa
618
11
Mtl Data To Text
Apache-2.0
The MTL-data-to-text model is specifically designed for data-to-text generation tasks, utilizing supervised pre-training on mixed-annotated data-to-text datasets.
Text Generation Transformers Supports Multiple Languages
M
RUCAIBox
147
28
Mtl Summarization
Apache-2.0
The MTL-Summarization Model is a text generation model specifically designed for summarization tasks, trained through supervised pre-training on multiple annotated summarization datasets.
Text Generation Transformers Supports Multiple Languages
M
RUCAIBox
13
0
T5 Arabic Base
Apache-2.0
This is a T5 model customized for Arabic and English tasks, serving as an alternative to the google/mt5-base model, with a smaller size and a focus on Arabic and English-related tasks.
Large Language Model Transformers
T
bakrianoo
635
0
S BioBert Snli Multinli Stsb
This is a model based on sentence-transformers that can map sentences and paragraphs into a 768-dimensional dense vector space, suitable for tasks such as sentence similarity calculation and semantic search.
Text Embedding Transformers
S
pritamdeka
987
7
T5 Arabic Small
Apache-2.0
A T5 model customized for Arabic and English tasks, smaller in size and optimized for specific languages.
Large Language Model Transformers
T
bakrianoo
20
1
Rubert Base Cased
RuBERT is a Russian BERT model trained on Russian Wikipedia and news data, with 180 million parameters, supporting masked language modeling and next sentence prediction tasks.
Large Language Model Other
R
DeepPavlov
275.78k
106
Code Trans T5 Base Source Code Summarization Python Multitask
A pre-trained model based on the T5 architecture, specifically designed for Python code summarization with multi-task support
Large Language Model
C
SEBIS
57
1
Code Trans T5 Large Code Documentation Generation Python Multitask
Python code documentation generation model based on T5-large architecture, supporting multi-task processing
Text Generation
C
SEBIS
91
5
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase